# 128K long text processing
Phi 3.5 Mini Instruct
MIT
Phi-3.5-mini-instruct is a lightweight and advanced open-source model built on the dataset used by Phi-3, focusing on high-quality, inference-rich data. It supports a 128K token context length and has powerful multilingual and long-context processing capabilities.
Large Language Model
Transformers Other

P
Lexius
129
1
Typhoon2.1 Gemma3 12b Gguf
Typhoon2.1-Gemma3-12B is an instruction-based Thai large language model with 12 billion parameters, a context length of 128K, and function call ability.
Large Language Model
T
scb10x
186
1
Typhoon2.1 Gemma3 4b Gguf
Typhoon2.1-Gemma3-4B is an instruction-based large Thai language model with 4 billion parameters and a context length of 128K, supporting function call ability.
Large Language Model
T
scb10x
218
2
Qwen2.5 7B Instruct
Apache-2.0
Qwen2.5 is the latest series of Tongyi large language models, offering a variety of models with parameters ranging from 500 million to 72 billion, with significant improvements in coding, math, instruction following, and long text processing.
Large Language Model
Transformers Supports Multiple Languages

Q
Gensyn
547.02k
2
Qwen3 32B 128K GGUF
Apache-2.0
Qwen3 is the latest generation of large language models in the Qwen series, offering a range of dense and mixture-of-experts (MoE) models. Based on extensive training, Qwen3 has made breakthroughs in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model English
Q
unsloth
20.51k
20
Google Gemma 3 27b It Text
Plain text adaptation of Gemma 3 27B, removing image processing capabilities to focus on text generation tasks
Large Language Model
Transformers

G
Changgil
34
2
Qwen2.5 7B Instruct GGUF Llamafile
Apache-2.0
Qwen2.5 is the latest series of the Tongyi Qianwen large model, including base models and instruction-tuned models with parameter scales ranging from 0.5B to 72B, showing significant improvements in areas such as code, mathematics, instruction following, and long text generation.
Large Language Model English
Q
Bojun-Feng
441
2
Gemma 3 1b It Qat
Gemma 3 is a lightweight multimodal model launched by Google, capable of processing text and image inputs and generating text outputs. This model has a 128K large context window and multilingual support for over 140 languages.
Image-to-Text
Transformers

G
unsloth
2,558
1
Gemma 3 1b It Llamafile
Gemma is a lightweight open model series launched by Google, built on the same research technology as Gemini. The llamafile version is packaged as an executable file by Mozilla for easy use on multiple platforms.
Text-to-Image
G
Mozilla
469
3
Deepseek R1 FP4
MIT
FP4 quantized version of DeepSeek R1 model, using optimized Transformer architecture for efficient text generation
Large Language Model
D
nvidia
61.51k
239
C4ai Command R Plus 08 2024
Command R+ 08-2024 is a 104 billion parameter multilingual large language model that supports Retrieval-Augmented Generation (RAG) and tool usage, suitable for automating complex tasks.
Large Language Model
Transformers Supports Multiple Languages

C
CohereLabs
4,265
260
C4ai Command R Plus 4bit
Cohere Labs Command R+ is a multilingual large language model with 104 billion parameters, featuring advanced capabilities such as retrieval-augmented generation (RAG) and tool usage, and supporting a context length of 128K.
Large Language Model
Transformers Supports Multiple Languages

C
CohereLabs
316
252
Featured Recommended AI Models